40 research outputs found

    The Science of Detecting LLM-Generated Texts

    Full text link
    The emergence of large language models (LLMs) has resulted in the production of LLM-generated texts that is highly sophisticated and almost indistinguishable from texts written by humans. However, this has also sparked concerns about the potential misuse of such texts, such as spreading misinformation and causing disruptions in the education system. Although many detection approaches have been proposed, a comprehensive understanding of the achievements and challenges is still lacking. This survey aims to provide an overview of existing LLM-generated text detection techniques and enhance the control and regulation of language generation models. Furthermore, we emphasize crucial considerations for future research, including the development of comprehensive evaluation metrics and the threat posed by open-source LLMs, to drive progress in the area of LLM-generated text detection

    SPeC: A Soft Prompt-Based Calibration on Mitigating Performance Variability in Clinical Notes Summarization

    Full text link
    Electronic health records (EHRs) store an extensive array of patient information, encompassing medical histories, diagnoses, treatments, and test outcomes. These records are crucial for enabling healthcare providers to make well-informed decisions regarding patient care. Summarizing clinical notes further assists healthcare professionals in pinpointing potential health risks and making better-informed decisions. This process contributes to reducing errors and enhancing patient outcomes by ensuring providers have access to the most pertinent and current patient data. Recent research has shown that incorporating prompts with large language models (LLMs) substantially boosts the efficacy of summarization tasks. However, we show that this approach also leads to increased output variance, resulting in notably divergent outputs even when prompts share similar meanings. To tackle this challenge, we introduce a model-agnostic Soft Prompt-Based Calibration (SPeC) pipeline that employs soft prompts to diminish variance while preserving the advantages of prompt-based summarization. Experimental findings on multiple clinical note tasks and LLMs indicate that our method not only bolsters performance but also effectively curbs variance for various LLMs, providing a more uniform and dependable solution for summarizing vital medical information

    DISPEL: Domain Generalization via Domain-Specific Liberating

    Full text link
    Domain generalization aims to learn a generalization model that can perform well on unseen test domains by only training on limited source domains. However, existing domain generalization approaches often bring in prediction-irrelevant noise or require the collection of domain labels. To address these challenges, we consider the domain generalization problem from a different perspective by categorizing underlying feature groups into domain-shared and domain-specific features. Nevertheless, the domain-specific features are difficult to be identified and distinguished from the input data. In this work, we propose DomaIn-SPEcific Liberating (DISPEL), a post-processing fine-grained masking approach that can filter out undefined and indistinguishable domain-specific features in the embedding space. Specifically, DISPEL utilizes a mask generator that produces a unique mask for each input data to filter domain-specific features. The DISPEL framework is highly flexible to be applied to any fine-tuned models. We derive a generalization error bound to guarantee the generalization performance by optimizing a designed objective loss. The experimental results on five benchmarks demonstrate DISPEL outperforms existing methods and can further generalize various algorithms

    Efficient XAI Techniques: A Taxonomic Survey

    Full text link
    Recently, there has been a growing demand for the deployment of Explainable Artificial Intelligence (XAI) algorithms in real-world applications. However, traditional XAI methods typically suffer from a high computational complexity problem, which discourages the deployment of real-time systems to meet the time-demanding requirements of real-world scenarios. Although many approaches have been proposed to improve the efficiency of XAI methods, a comprehensive understanding of the achievements and challenges is still needed. To this end, in this paper we provide a review of efficient XAI. Specifically, we categorize existing techniques of XAI acceleration into efficient non-amortized and efficient amortized methods. The efficient non-amortized methods focus on data-centric or model-centric acceleration upon each individual instance. In contrast, amortized methods focus on learning a unified distribution of model explanations, following the predictive, generative, or reinforcement frameworks, to rapidly derive multiple model explanations. We also analyze the limitations of an efficient XAI pipeline from the perspectives of the training phase, the deployment phase, and the use scenarios. Finally, we summarize the challenges of deploying XAI acceleration methods to real-world scenarios, overcoming the trade-off between faithfulness and efficiency, and the selection of different acceleration methods.Comment: 15 pages, 3 figure

    Towards Assumption-free Bias Mitigation

    Full text link
    Despite the impressive prediction ability, machine learning models show discrimination towards certain demographics and suffer from unfair prediction behaviors. To alleviate the discrimination, extensive studies focus on eliminating the unequal distribution of sensitive attributes via multiple approaches. However, due to privacy concerns, sensitive attributes are often either unavailable or missing in real-world scenarios. Therefore, several existing works alleviate the bias without sensitive attributes. Those studies face challenges, either in inaccurate predictions of sensitive attributes or the need to mitigate unequal distribution of manually defined non-sensitive attributes related to bias. The latter requires strong assumptions about the correlation between sensitive and non-sensitive attributes. As data distribution and task goals vary, the strong assumption on non-sensitive attributes may not be valid and require domain expertise. In this work, we propose an assumption-free framework to detect the related attributes automatically by modeling feature interaction for bias mitigation. The proposed framework aims to mitigate the unfair impact of identified biased feature interactions. Experimental results on four real-world datasets demonstrate that our proposed framework can significantly alleviate unfair prediction behaviors by considering biased feature interactions

    DiscoverPath: A Knowledge Refinement and Retrieval System for Interdisciplinarity on Biomedical Research

    Full text link
    The exponential growth in scholarly publications necessitates advanced tools for efficient article retrieval, especially in interdisciplinary fields where diverse terminologies are used to describe similar research. Traditional keyword-based search engines often fall short in assisting users who may not be familiar with specific terminologies. To address this, we present a knowledge graph-based paper search engine for biomedical research to enhance the user experience in discovering relevant queries and articles. The system, dubbed DiscoverPath, employs Named Entity Recognition (NER) and part-of-speech (POS) tagging to extract terminologies and relationships from article abstracts to create a KG. To reduce information overload, DiscoverPath presents users with a focused subgraph containing the queried entity and its neighboring nodes and incorporates a query recommendation system, enabling users to iteratively refine their queries. The system is equipped with an accessible Graphical User Interface that provides an intuitive visualization of the KG, query recommendations, and detailed article information, enabling efficient article retrieval, thus fostering interdisciplinary knowledge exploration. DiscoverPath is open-sourced at https://github.com/ynchuang/DiscoverPath

    Recent work on sprite spectrum in Taiwan

    Full text link
    campaigns in Taiwan. We first introduce two types of spectroimagers, the slit and slitless types, and discuss their advantages and shortcomings. Next we explore the instrument development and procedures undertaken for this study. In 2006, a slit spectroimager was installed for a sprite campaign and on 15 August of that year, two sprite spectra were recorded using the slit spectroimager along with seven sprites, one halo, one ELVES emission and two jets. By the end of 2015, a slitless spectroimager had been successfully constructed and was ready to conduct additional investigations. On 7 May 2016, a sprite spectrum was recorded using the slitless spectroimager. Following an examination of the calibrations (comprising detection region field of view, wavelength calibration, and response curve), data analysis, and additional calibrations (comprising elevation and azimuthal angles, atmospheric transmittance, and theoretical wavelength calculations) performed in this study, we present the results from our observed sprite spectra using the slit and slitless spectroimagers

    An Overview of Regional Experiments on Biomass Burning Aerosols and Related Pollutants in Southeast Asia: From BASE-ASIA and the Dongsha Experiment to 7-SEAS

    Get PDF
    By modulating the Earth-atmosphere energy, hydrological and biogeochemical cycles, and affecting regional-to-global weather and climate, biomass burning is recognized as one of the major factors affecting the global carbon cycle. However, few comprehensive and wide-ranging experiments have been conducted to characterize biomass-burning pollutants in Southeast Asia (SEA) or assess their regional impact on meteorology, the hydrological cycle, the radiative budget, or climate change. Recently, BASEASIA (Biomass-burning Aerosols in South-East Asia: Smoke Impact Assessment) and the 7-SEAS (7- South-East Asian Studies) Dongsha Experiment were conducted during the spring seasons of 2006 and 2010 in northern SEA, respectively, to characterize the chemical, physical, and radiative properties of biomass-burning emissions near the source regions, and assess their effects. This paper provides an overview of results from these two campaigns and related studies collected in this special issue, entitled Observation, modeling and impact studies of biomass burning and pollution in the SE Asian Environment. This volume includes 28 papers, which provide a synopsis of the experiments, regional weatherclimate, chemical characterization of biomass-burning aerosols and related pollutants in source and sink regions, the spatial distribution of air toxics (atmospheric mercury and dioxins) in source and remote areas, a characterization of aerosol physical, optical, and radiative properties, as well as modeling and impact studies. These studies, taken together, provide the first relatively complete dataset of aerosol chemistry and physical observations conducted in the sourcesink region in the northern SEA, with particular emphasis on the marine boundary layer and lower free troposphere (LFT). The data, analysis and modeling included in these papers advance our present knowledge of source characterization of biomass-burning pollutants near the source regions as well as the physical and chemical processes along transport pathways. In addition, we raise key questions to be addressed by a coming deployment during springtime 2013 in northern SEA, named 7-SEASBASELInE (Biomass-burning Aerosols Stratocumulus Environment: Lifecycles and Interactions Experiment). This campaign will include a synergistic approach for further exploring many key atmospheric processes (e.g., complex aerosol-cloud interactions) and impacts of biomass burning on the surface-atmosphere energy budgets during the lifecycles of biomass burning emissions
    corecore